Picture for Wenjuan Han

Wenjuan Han

SToLa: Self-Adaptive Touch-Language Framework with Tactile Commonsense Reasoning in Open-Ended Scenarios

Add code
May 07, 2025
Viaarxiv icon

Extralonger: Toward a Unified Perspective of Spatial-Temporal Factors for Extra-Long-Term Traffic Forecasting

Add code
Oct 30, 2024
Viaarxiv icon

Touch100k: A Large-Scale Touch-Language-Vision Dataset for Touch-Centric Multimodal Representation

Add code
Jun 06, 2024
Figure 1 for Touch100k: A Large-Scale Touch-Language-Vision Dataset for Touch-Centric Multimodal Representation
Figure 2 for Touch100k: A Large-Scale Touch-Language-Vision Dataset for Touch-Centric Multimodal Representation
Figure 3 for Touch100k: A Large-Scale Touch-Language-Vision Dataset for Touch-Centric Multimodal Representation
Figure 4 for Touch100k: A Large-Scale Touch-Language-Vision Dataset for Touch-Centric Multimodal Representation
Viaarxiv icon

Transformer in Touch: A Survey

Add code
May 21, 2024
Viaarxiv icon

Potential and Limitations of LLMs in Capturing Structured Semantics: A Case Study on SRL

Add code
May 10, 2024
Viaarxiv icon

MineLand: Simulating Large-Scale Multi-Agent Interactions with Limited Multimodal Senses and Physical Needs

Add code
Mar 28, 2024
Figure 1 for MineLand: Simulating Large-Scale Multi-Agent Interactions with Limited Multimodal Senses and Physical Needs
Figure 2 for MineLand: Simulating Large-Scale Multi-Agent Interactions with Limited Multimodal Senses and Physical Needs
Figure 3 for MineLand: Simulating Large-Scale Multi-Agent Interactions with Limited Multimodal Senses and Physical Needs
Figure 4 for MineLand: Simulating Large-Scale Multi-Agent Interactions with Limited Multimodal Senses and Physical Needs
Viaarxiv icon

Towards Comprehensive Multimodal Perception: Introducing the Touch-Language-Vision Dataset

Add code
Mar 14, 2024
Figure 1 for Towards Comprehensive Multimodal Perception: Introducing the Touch-Language-Vision Dataset
Figure 2 for Towards Comprehensive Multimodal Perception: Introducing the Touch-Language-Vision Dataset
Figure 3 for Towards Comprehensive Multimodal Perception: Introducing the Touch-Language-Vision Dataset
Figure 4 for Towards Comprehensive Multimodal Perception: Introducing the Touch-Language-Vision Dataset
Viaarxiv icon

TransGPT: Multi-modal Generative Pre-trained Transformer for Transportation

Add code
Feb 11, 2024
Figure 1 for TransGPT: Multi-modal Generative Pre-trained Transformer for Transportation
Figure 2 for TransGPT: Multi-modal Generative Pre-trained Transformer for Transportation
Figure 3 for TransGPT: Multi-modal Generative Pre-trained Transformer for Transportation
Figure 4 for TransGPT: Multi-modal Generative Pre-trained Transformer for Transportation
Viaarxiv icon

TransportationGames: Benchmarking Transportation Knowledge of (Multimodal) Large Language Models

Add code
Jan 09, 2024
Viaarxiv icon

CLOVA: A Closed-Loop Visual Assistant with Tool Usage and Update

Add code
Dec 18, 2023
Viaarxiv icon